Recently, a new evaluation benchmark - RBench-V, specifically designed to test the visual reasoning capabilities of multi-modal large models, was released by research teams from Tsinghua University, Tencent HUNYUAN, Stanford University, and Carnegie Mellon University. The introduction of this benchmark aims to fill the gap in the current evaluation system regarding the model's visual output capabilities, allowing for a more comprehensive understanding of existing model performance. The RBench-V benchmark consists of 803 questions covering multiple fields, including geometry and graph theory, mechanics and electromagnetism, multi-target recognition, and path planning.